PAC-Learning Geometrical Figures
نویسنده
چکیده
The thesis studies the following problem: Given a set of geometrical figures (such as planar polygons), each one labelled according to whether or not it resembles some "ideal" figure, find a good approximation to that ideal figure which can be used to classify other figures in the same way. We work within the PAC learning model introduced by Valiant in 1984. Informally, the concepts under consideration are sets of polygons which resemble each other visually. A learning algorithm is given collections of members and non-members of a concept, and its task is to infer a criterion for membership which is consistent with the given examples and which can be used as an accurate classifier of further example polygons. In order, to formalise the notion of a concept, we use metrics which measure the extent to which two polygons differ. A concept is assumed to be the set of polygons which are within some distance of some fixed central polygon. In the thesis we work most extensively with the Hausdorif metric. Using the Hausdorif metric we obtain NP-completeness results for several variants of the learning problem. In particular we show that it is hard to find a single geometrical figure which is close to the positive examples but not to the negative examples. This result holds under various assumptions about the specific geometrical figures under consideration. It also holds for several metrics other than the Hausdorff metric. Despite the NP-completeness results mentioned above we have found some encouraging positive results. In particular, we have discovered a general technique for prediction. (Prediction is a less demanding learning model thah. PAC learning. The goal is to find a polynomial-time algorithm which takes as input a sample of labelled examples and is then able to predict the status of further unlabelled examples in polynomial time.) Using our technique we have obtained polynomial-time algorithms for predicting many of the geometrical concept classes studied in the thesis. These algorithms do not classify geometrical figures by measuring their distance from a single "ideal" geometrical figure. Instead, they identify a collection of concepts whose intersection may be used to classify examples reliably. It is natural to consider the case in which only positive examples are available. In the thesis we show that some but not all of the concept classes may be
منابع مشابه
Minimizing Disagreements for Geometric Regions Using Dynamic Programming, with Applications to Machine Learning and Computer Graphics
We demonstrate that the dynamic programming paradigm is an eeective tool in the design of eecient algorithms for solving minimumdisagreement problems for convex polygons, star-shaped polygons, unions of axis-parallel boxes and various other classes of geometric regions. In particular, we show that the minimizing disagreement problem for convex k-gons on a sample of size n can be solved in O(n 6...
متن کاملGeometrical Intuition and the Learning and Teaching of Geometry
Intuition is often regarded as essential in the learning of geometry, but how such skills might be effectively developed in students remains an open question. This paper reviews the role and importance of geometrical intuition and suggests it involves the skills to create and manipulate geometrical figures in the mind, to see geometrical properties, to relate images to concepts and theorems in ...
متن کاملSimplification of Fan-Meshes Models for Fast Rendering of Large 3D Point-Sampled Scenes
Fan-Meshes (FM) are a kind of geometrical primitives for generating 3D model or scene descriptions that are able to preserve both local geometrical details and topological structures. In this paper, we propose an efficient simplification algorithm for the FM models to achieve fast post-processing and rendering of large models or scenes. Given a global error tolerance for the surface approximati...
متن کاملComputational Learning Theory Fall Semester , 2010 Lecture 3 : October 31
In this lecture we will talk about the PAC model. The PAC learning model is one of the important and famous learning model. PAC stands for Probably Approximately Correct, our goal is to learn a hypothesis from a hypothesis class such that in high con dence we will have a small error rate (approximately correct). We start the lecture with an intuitive example to explain the idea behind the PAC m...
متن کاملA Pac{bayesian Study of Linear Classiiers: Why Svms Work
In this paper we present a bound on the generalisation error of linear classi ers in terms of a re ned margin quantity on the training set. The result is obtained by a fundamentally di erent reasoning than in the classical PAC framework. We show that the generalisation error of a classi er can be bounded purely by geometrical arguments. In contrast to the classical results our bound does not co...
متن کامل